Usage and examples
Starting the Demo Application
If it is the first time you will run the demo, please start in step 1, if not, you can skip steps 1, 2, and 3 and start in step 4.
1. The first step is to get the demo source code, do it by running:
git clone https://github.com/RidgeRun/smart-seek-360.git
The content of the repository should look as follows:
. ├── config │ ├── agent │ │ ├── api_mapping.json │ │ └── prompt.txt │ ├── analytics │ │ └── configuration.json │ ├── demo-nginx.conf │ └── vst │ ├── vst_config.json │ └── vst_storage.json ├── docker-compose.yaml └── README.md
2. Update Ingress Configuration
cd smart-seek-360 sudo cp config/demo-nginx.conf /opt/nvidia/jetson/services/ingress/config/
3. Update VST Configuration
cd smart-seek-360 sudo cp config/vst/* /opt/nvidia/jetson/services/vst/config/
4. Launch Platform Services
sudo systemctl start jetson-redis sudo systemctl start jetson-ingress sudo systemctl start jetson-vst
5. Add at least one stream to VST. You can follow VST documentation for instructions on how to do it.
6. Launch Smart Seek 360
cd smart-seek-360 docker compose up -d
Demo Usage
If you followed the previous steps to start the application, you should have a stream available at rtsp://BOARD_IP:5021/ptz_out. You can open it from the host computer with the following command:
vlc rtsp://BOARD_IP:5021/ptz_out
Just replace BOARD_IP with the actual IP address of the board running the demo.
The demo is controlled through the AI-Agent which comes with a web interface that can be accessed at BOART_IP:30080/agent.
The page looks as follows:
Through that interface, the application can be controlled using natural language, the 2 available commands are: move camera and find objects.
Move Camera
The available options are:
- move the camera X degrees left: This will move the camera X degrees specified to the left.
- move the camera X degrees right: This will move the camera X degrees specified to the right.
Find Objects
With this feature, you can indicate the application to look for any object on the input stream and 2 actions will be performed.
1. The camera will point to that object once it is found.
2. A clip will be recorded of that event (disabled by default).
Demo in action
The following video shows how to start and run the demo.